Mind2Mind: Transfer Learning for GANs
نویسندگان
چکیده
Training generative adversarial networks (GANs) on high quality (HQ) images involves important computing resources. This requirement represents a bottleneck for the development of applications GANs. We propose transfer learning technique GANs that significantly reduces training time. Our approach consists freezing low-level layers both critic and generator original GAN. assume an auto-encoder constraint to ensure compatibility internal representations generator. assumption explains gain in time as it enables us bypass during forward backward passes. compare our method baselines observe significant acceleration training. It can reach two orders magnitude HQ datasets when compared with StyleGAN. provide theorem, rigorously proven within framework optimal transport, ensuring convergence transferred moreover precise bound terms distance between source target dataset.
منابع مشابه
Memorization Precedes Generation: Learning Unsupervised Gans
We propose an approach to address two undesired properties of unsupervised GANs. First, since GANs use only a continuous latent distribution to embed multiple classes or clusters of a dataset, GANs often do not correctly handle the structural discontinuity between disparate classes in a latent space. Second, discriminators of GANs easily forget about past generated samples by generators, incurr...
متن کاملMemorization Precedes Generation: Learning Unsupervised Gans
We propose an approach to address two undesired properties of unsupervised GANs. First, since GANs use only a continuous latent distribution to embed multiple classes or clusters of a dataset, GANs often do not correctly handle the structural discontinuity between disparate classes in a latent space. Second, discriminators of GANs easily forget about past generated samples by generators, incurr...
متن کاملMemorization Precedes Generation: Learning Unsupervised Gans
We propose an approach to address two undesired properties of unsupervised GANs. First, since GANs use only a continuous latent distribution to embed multiple classes or clusters of a dataset, GANs often do not correctly handle the structural discontinuity between disparate classes in a latent space. Second, discriminators of GANs easily forget about past generated samples by generators, incurr...
متن کاملHierarchical Functional Concepts for Knowledge Transfer among Reinforcement Learning Agents
This article introduces the notions of functional space and concept as a way of knowledge representation and abstraction for Reinforcement Learning agents. These definitions are used as a tool of knowledge transfer among agents. The agents are assumed to be heterogeneous; they have different state spaces but share a same dynamic, reward and action space. In other words, the agents are assumed t...
متن کاملSemi-supervised Learning with GANs: Manifold Invariance with Improved Inference
Semi-supervised learning methods using Generative Adversarial Networks (GANs) have shown promising empirical success recently. Most of these methods use a shared discriminator/classifier which discriminates real examples from fake while also predicting the class label. Motivated by the ability of the GANs generator to capture the data manifold well, we propose to estimate the tangent space to t...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Lecture Notes in Computer Science
سال: 2021
ISSN: ['1611-3349', '0302-9743']
DOI: https://doi.org/10.1007/978-3-030-80209-7_91